92¢ /MFlops/s, Ultra-Large-Scale Neural-Network Training on a PIII Cluster

نویسنده

  • Gordon Bell
چکیده

Artificial neural networks with millions of adjustable parameters and a similar number of training examples are a potential solution for difficult, large-scale pattern recognition problems in areas such as speech and face recognition, classification of large volumes of web data, and finance. The bottleneck is that neural network training involves iterative gradient descent and is extremely computationally intensive. In this paper we present a technique for distributed training of Ultra Large Scale Neural Networks1 (ULSNN) on Bunyip, a Linux-based cluster of 196 Pentium III processors. To illustrate ULSNN training we describe an experiment in which a neural network with 1.73 million adjustable parameters was trained to recognize machineprinted Japanese characters from a database containing 9 million training patterns. The training runs with a average performance of 163.3 GFlops/s (single precision). With a machine cost of $150,913, this yields a price/performance ratio of 92.4¢ /MFlops/s (single precision). For comparison purposes, training using double precision and the ATLAS DGEMM produces a sustained performance of 70 MFlops/s or $2.16 / MFlop/s (double precision).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

98¢/Mflops/s, Ultra-Large-Scale Neural-Network Training on a PIII Cluster

Artificial neural networks with millions of adjustable parameters and a similar number of training examples are a potential solution for difficult, large-scale pattern recognition problems in areas such as speech and face recognition, classification of large volumes of web data, and finance. The bottleneck is that neural network training involves iterative gradient descent and is extremely comp...

متن کامل

98¢ /MFlop, Ultra-Large-Scale Neural-Network Training on a PIII Cluster

Artificial neural networks with millions of adjustable parameters and a similar number of training examples are a potential solution for difficult, large-scale pattern recognition problems in areas such as speech and face recognition, classification of large volumes of web data, and finance. The bottleneck is that neural network training involves iterative gradient descent and is extremely comp...

متن کامل

پیش‌‌بینی کوتاه مدت قیمت تراکم گرهی در یک سیستم قدرت بزرگ تجدید ساختار یافته با استفاده از شبکه‌های عصبی مصنوعی با بهینه‌سازی آموزش ژنتیکی

In a daily power market, price and load forecasting is the most important signal for the market participants. In this paper, an accurate feed-forward neural network model with a genetic optimization levenberg-marquardt back propagation (LMBP) training algorithm is employed for short-term nodal congestion price forecasting in different zones of a large-scale power market. The use of genetic algo...

متن کامل

Link Prediction using Network Embedding based on Global Similarity

Background: The link prediction issue is one of the most widely used problems in complex network analysis. Link prediction requires knowing the background of previous link connections and combining them with available information. The link prediction local approaches with node structure objectives are fast in case of speed but are not accurate enough. On the other hand, the global link predicti...

متن کامل

Implementation of Parallelizing Multi-layer Neural Networks Based on Cloud Computing

Background: Cloud computing, as a technology developed under the rapid development of modern network, is mainly used for processing large-scale data. The traditional data mining algorithms such as neural network algorithm are usually used for processing small-scale data. Therefore, the calculation of large-scale data using neural network algorithm must be based on cloud computing. Materials and...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000